A Neural Autoregressive Topic Model
نویسندگان
چکیده
We describe a new model for learning meaningful representations of text documents from an unlabeled collection of documents. This model is inspired by the recently proposed Replicated Softmax, an undirected graphical model of word counts that was shown to learn a better generative model and more meaningful document representations. Specifically, we take inspiration from the conditional mean-field recursive equations of the Replicated Softmax in order to define a neural network architecture that estimates the probability of observing a new word in a given document given the previously observed words. This paradigm also allows us to replace the expensive softmax distribution over words with a hierarchical distribution over paths in a binary tree of words. The end result is a model whose training complexity scales logarithmically with the vocabulary size instead of linearly as in the Replicated Softmax. Our experiments show that our model is competitive both as a generative model of documents and as a document representation learning algorithm.
منابع مشابه
A Nonlinear Autoregressive Model with Exogenous Variables Neural Network for Stock Market Timing: The Candlestick Technical Analysis
In this paper, the nonlinear autoregressive model with exogenous variables as a new neural network is used for timing of the stock markets on the basis of the technical analysis of Japanese Candlestick. In this model, the “nonlinear autoregressive model with exogenous variables” is an analyzer. For a more reliable comparison, here (like the literature) two approaches of Raw-based and Signal-ba...
متن کاملA Supervised Neural Autoregressive Topic Model for Simultaneous Image Classification and Annotation
Topic modeling based on latent Dirichlet allocation (LDA) has been a framework of choice to perform scene recognition and annotation. Recently, a new type of topic model called the Document Neural Autoregressive Distribution Estimator (DocNADE) was proposed and demonstrated state-of-the-art performance for document modeling. In this work, we show how to successfully apply and extend this model ...
متن کاملDocument Neural Autoregressive Distribution Estimation
We present an approach based on feed-forward neural networks for learning the distribution of textual documents. This approach is inspired by the Neural Autoregressive Distribution Estimator (NADE) model, which has been shown to be a good estimator of the distribution of discrete-valued high-dimensional vectors. In this paper, we present how NADE can successfully be adapted to the case of textu...
متن کاملPrediction of Above-elbow Motions in Amputees, based on Electromyographic(EMG) Signals, Using Nonlinear Autoregressive Exogenous (NARX) Model
Introduction In order to improve the quality of life of amputees, biomechatronic researchers and biomedical engineers have been trying to use a combination of various techniques to provide suitable rehabilitation systems. Diverse biomedical signals, acquired from a specialized organ or cell system, e.g., the nervous system, are the driving force for the whole system. Electromyography(EMG), as a...
متن کاملComparison of Neural Network Models, Vector Auto Regression (VAR), Bayesian Vector-Autoregressive (BVAR), Generalized Auto Regressive Conditional Heteroskedasticity (GARCH) Process and Time Series in Forecasting Inflation in Iran
This paper has two aims. The first is forecasting inflation in Iran using Macroeconomic variables data in Iran (Inflation rate, liquidity, GDP, prices of imported goods and exchange rates) , and the second is comparing the performance of forecasting vector auto regression (VAR), Bayesian Vector-Autoregressive (BVAR), GARCH, time series and neural network models by which Iran's inflation is for...
متن کاملFast Decoding in Sequence Models Using Discrete Latent Variables
Autoregressive sequence models based on deep neural networks, such as RNNs, Wavenet and the Transformer attain state-of-the-art results on many tasks. However, they are difficult to parallelize and are thus slow at processing long sequences. RNNs lack parallelism both during training and decoding, while architectures like WaveNet and Transformer are much more parallelizable during training, yet...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2012